Toward Adaptive Information Fusion in Multimodal Systems

نویسندگان

  • Xiao Huang
  • Sharon L. Oviatt
چکیده

Techniques for information fusion are at the heart of multimodal system design. To develop new user-adaptive approaches for multimodal fusion, our lab has investigated the stability and basis of major individual differences that have been documented in users’ multimodal integration patterns. In this talk, I summarized the following: (1) there are large individual differences in users’ dominant speech and pen multimodal integration pattern, such that individual users can be classified as either simultaneous or sequential integrators (Oviatt, 1999; Oviatt et al., 2003), (2) users’ dominant integration pattern can be identified almost immediately (i.e., upon first interaction with computer), and it remains highly consistent over a session (Oviatt et al., 2003; Oviatt et al., 2005b), (3) users’ dominant integration pattern also remains stable across their lifespan (Oviatt et al., 2003; Oviatt et al., 2005b), (4) users’ dominant integration pattern is highly resistant to change, even when they are given strong selective reinforcement or explicit instructions to switch patterns (Oviatt et al., 2003; Oviatt et al., 2005a), (5) when users encounter cognitive load (e.g., due to increasing task difficulty, or system recognition errors), their dominant multimodal integration pattern entrenches or becomes “hypertimed,” (Oviatt et al., 2003; Oviatt et al., 2004), and (6) users’ distinctive integration patterns appear to derive from enduring differences in basic reflective-impulsive cognitive style (Oviatt et al., 2005b). In this talk, I also discussed recent work in our lab that combines empirical user modeling with machine learning techniques to learn users’ multimodal integration patterns. This work emphasizes the establishment of user-adaptive temporal thresholds for time-critical multimodal systems, rather than fixed temporal thresholds which are the current state-of-the-art. Estimates indicate that system delays can be reduced to just 44% of what they currently Copyright © 2006, Australian Computer Society, Inc. This paper appeared at the NICTA-HCSNet Multimodal User Interaction Workshop (MMUI2005), Sydney, Australia. Conferences in Research and Practice in Information Technology, Vol. 57. Fang Chen and Julien Epps, Eds. Reproduction for academic, not-for profit purposes permitted provided this text is included. are by adopting user-defined thresholds, with related substantial reductions in system recognition errors. Ongoing research in our group is exploring which machine learning techniques and models provide the best acceleration and generalization of learned multimodal integration patterns, reliability of signal and information fusion, and overall improvement in multimodal interpretation. We currently are developing a three-tiered user-adaptive model to conduct on-line adaptation of a multimodal system’s temporal thresholds during fusion based on a user’s habitual integration pattern, which is used as prior knowledge. Implications of this research were discussed for the design of next-generation adaptive multimodal systems with substantially improved performance characteristics.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Multimodal medical image fusion based on Yager’s intuitionistic fuzzy sets

The objective of image fusion for medical images is to combine multiple images obtained from various sources into a single image suitable for better diagnosis. Most of the state-of-the-art image fusing technique is based on nonfuzzy sets, and the fused image so obtained lags with complementary information. Intuitionistic fuzzy sets (IFS) are determined to be more suitable for civilian, and medi...

متن کامل

Adaptive management of multimodal biometrics fusion using ant colony optimization

This paper presents a new approach for the adaptive management of multimodal biometrics to meet a wide range of application dependent adaptive security requirements. In this work, ant colony optimization (ACO) is employed for the selection of key parameters like decision threshold and fusion rule, to ensure the optimal performance in meeting varying security requirements during the deployment o...

متن کامل

Adaptive Multimodal Fusion

Multimodal interfaces offer its users the possibility of interacting with computers, in a transparent, natural way, by means of various modalities. Fusion engines are key components in multimodal systems, responsible for combining information from different sources and extract a semantic meaning from them. This fusion process allows many modalities to be effectively used at once and therefore a...

متن کامل

An Adaptive Approach to Collecting Multimodal Input

Multimodal dialogue systems allow users to input information in multiple modalities. These systems can handle simultaneous or sequential composite multimodal input. Different coordination schemes require such systems to capture, collect and integrate user input in different modalities, and then respond to a joint interpretation. We performed a study to understand the variability of input in mul...

متن کامل

Knowledge Based Multimodal Result Fusion for Distributed and Heterogeneous Multimedia Environments: Concept and Ideas

Distributed multimedia retrieval (DMR) is a key issue in today’s information systems. One problem in DMR is the fusion of results retrieved from multiple locations, which is required in order to present the results in an integrated, consolidated and aligned form. This paper sketches a distributed multimedia retrieval scenario in which requirements, a conceptualization and an implementation of a...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2005